Source-Word Decomposition for Neural Machine Translation
نویسندگان
چکیده
منابع مشابه
Word Decomposition for Machine Translation
All feasible systems of machine translation are based on a unit smaller, in a great many cases, than the word. This unit, which provides the source-language entries in a mechanical dictionary, is conveniently termed a “chunk” so as to avoid confusion with other linguistic categories. There are, however, a number of ways in which words may be decomposed for machine-translation purposes and the f...
متن کاملBridging Source and Target Word Embeddings for Neural Machine Translation
Neural machine translation systems encode a source sequence into a vector from which a target sequence is generated via a decoder. Different from the traditional statistical machine translation, source and target words are not directly mapped to each other in translation rules. They are at the two ends of a long information channel in the encoder-decoder neural network, separated by source and ...
متن کاملModeling Source Syntax for Neural Machine Translation
Even though a linguistics-free sequence to sequence model in neural machine translation (NMT) has certain capability of implicitly learning syntactic information of source sentences, this paper shows that source syntax can be explicitly incorporated into NMT effectively to provide further improvements. Specifically, we linearize parse trees of source sentences to obtain structural label sequenc...
متن کاملWord Representations in Factored Neural Machine Translation
Translation into a morphologically rich language requires a large output vocabulary to model various morphological phenomena, which is a challenge for neural machine translation architectures. To address this issue, the present paper investigates the impact of having two output factors with a system able to generate separately two distinct representations of the target words. Within this framew...
متن کاملEmbedding Word Similarity with Neural Machine Translation
Neural language models learn word representations, or embeddings, that capture rich linguistic and conceptual information. Here we investigate the embeddings learned by neural machine translation models, a recently-developed class of neural language model. We show that embeddings from translation models outperform those learned by monolingual models at tasks that require knowledge of both conce...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Mathematical Problems in Engineering
سال: 2020
ISSN: 1563-5147,1024-123X
DOI: 10.1155/2020/4795187